BackPropagation Through Time

نویسنده

  • Jiang Guo
چکیده

This report provides detailed description and necessary derivations for the BackPropagation Through Time (BPTT) algorithm. BPTT is often used to learn recurrent neural networks (RNN). Contrary to feed-forward neural networks, the RNN is characterized by the ability of encoding longer past information, thus very suitable for sequential models. The BPTT extends the ordinary BP algorithm to suit the recurrent neural architecture. 1 Basic Definitions For a two-layer feed-forward neural network, we notate the input layer as x indexed by variable i, the hidden layer as s indexed by variable j, and the output layer as y indexed by variable k. The weight matrix that map the input vector to the hidden layer is V, while the hidden layer is propagated through the weight matrix W, to the output layer. In a simple recurrent neural network, we attach every neural layer a time subscript t. The input layer consists of two components, x(t) and the privious activation of the hidden layer s(t − 1) indexed by variable h. The corresponding weight matrix is U. Table 1 lists all the notations used in this report: Neural layer Description Index variable x(t) input layer i s(t− 1) previous hidden (state) layer h s(t) hidden (state) layer j y(t) output layer k Weight matrix Description Index variables V Input layer → Hidden layer i, j U Previous hidden layer → Hidden layer h, j W Hidden layer → Output layer j, k Table 1: Notations in the recurrent neural network. Then, the recurrent neural network can be processed as the following: • Input layer → Hidden layer sj(t) = f(netj(t)) (1)

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Relating Real-Time Backpropagation and Backpropagation-Through-Time: An Application of Flow Graph Interreciprocity

We show that signal ow graph theory provides a simple way to relate two popular algorithms used for adapting dynamic neural networks, real-time backpropagation and backpropagation-through-time. Starting with the ow graph for real-time backpropagation, we use a simple transposition to produce a second graph. The new graph is shown to be interreciprocal with the original and to correspond to the ...

متن کامل

Relating Real - Time Backpropagation andBackpropagation - Through - Time : An Application of Flow

We show that signal ow graph theory provides a simple way to relate two popular algorithms used for adapting dynamic neural networks, real-time backpropagation and backpropagation-through-time. Starting with the ow graph for real-time backpropagation, we use a simple transposition to produce a second graph. The new graph is shown to be interreciprocal with the original and to correspond to the ...

متن کامل

Backpropagation Through Time with Fixed Memory Size Requirements

and ei(t) is the output error, xi(t) represent the activations and δi(t) are the backpropagated errors. The system described by Eq. 1 and Eq. 2 constitute the backpropagation through time (BPTT) algorithm. Note that the backpropagation system (Eq. 2) should be run from t=T backwards to t=1. We define the boundary conditions δi(T+1)=0. We will assume that the instantaneous error signal ei(t) is ...

متن کامل

Time Series Forecasting Using Computational Intelligence Methods

Abstract: Forecasting the future evolution of a system based only on past information comprises a central scientific problem. In this work we investigate the comparative performance of recurrent multi–layer perceptrons, trained through backpropagation through time and the differential evolution algorithm, to perform one–step–ahead predictions for the laser time series (Data set A) from the Sant...

متن کامل

Backpropagation Through Time: What It Does and How to Do It

Backpropagation is now the most widely used tool in the field of artificial neural networks. At the core of backpropagation is a method for calculating derivatives exactly and efficiently in any large system made up of elementary subsystems or calculations which are represented by known, differentiable functions; thus, backpropagation has many applications which do not involve neural networks a...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2013